Bayesian L1-Norm Sparse Learning
نویسندگان
چکیده
We propose a Bayesian framework for learning the optimal regularization parameter in the L1-norm penalized least-mean-square (LMS) problem, also known as LASSO [1] or basis pursuit [2]. The setting of the regularization parameter is critical for deriving a correct solution. In most existing methods, the scalar regularization parameter is often determined in a heuristic manner; in contrast, our approach infers the optimal regularization setting under a Bayesian framework. Furthermore, Bayesian inference enables an independent regularization scheme where each coefficient (or weight) is associated with an independent regularization parameter. Simulations illustrate the improvement using our method in discovering sparse structure from noisy data. Comments Copyright 2006 IEEE. Reprinted from Proceedings of the 2006 IEEE International Conference Acoustics, Speech and Signal Processing, Volume 5, pages V605-V608. Publisher URL: http://dx.doi.org/10.1109/ ICASSP.2006.1661348 This material is posted here with permission of the IEEE. Such permission of the IEEE does not in any way imply IEEE endorsement of any of the University of Pennsylvania's products or services. Internal or personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution must be obtained from the IEEE by writing to [email protected]. By choosing to view this document, you agree to all provisions of the copyright laws protecting it. This conference paper is available at ScholarlyCommons: http://repository.upenn.edu/ese_papers/230 BAYESIAN L1-NORM SPARSE LEARNING Yuanqing Lin, Daniel D. Lee GRASP Laboratory, Department of Electrical and Systems Engineering, University of Pennsylvania, Philadelphia, PA 19104 ABSTRACT We propose a Bayesian framework for learning the optimal regularization parameter in the L1-norm penalized leastmean-square (LMS) problem, also known as LASSO [1] or basis pursuit [2]. The setting of the regularization parameter is critical for deriving a correct solution. In most existing methods, the scalar regularization parameter is often determined in a heuristic manner; in contrast, our approach infers the optimal regularization setting under a Bayesian framework. Furthermore, Bayesian inference enables an independent regularization scheme where each coefficient (or weight) is associated with an independent regularization parameter. Simulations illustrate the improvement using our method in discovering sparse structure from noisy data.We propose a Bayesian framework for learning the optimal regularization parameter in the L1-norm penalized leastmean-square (LMS) problem, also known as LASSO [1] or basis pursuit [2]. The setting of the regularization parameter is critical for deriving a correct solution. In most existing methods, the scalar regularization parameter is often determined in a heuristic manner; in contrast, our approach infers the optimal regularization setting under a Bayesian framework. Furthermore, Bayesian inference enables an independent regularization scheme where each coefficient (or weight) is associated with an independent regularization parameter. Simulations illustrate the improvement using our method in discovering sparse structure from noisy data.
منابع مشابه
Improved k-t FOCUSS using a sparse Bayesian learning
Introduction: In dynamic MRI, spatio-temporal resolution is a very important issue. Recently, compressed sensing approach has become a highly attracted imaging technique since it enables accelerated acquisition without aliasing artifacts. Our group has proposed an l1-norm based compressed sensing dynamic MRI called k-t FOCUSS which outperforms the existing methods. However, it is known that the...
متن کاملRobust Image Analysis by L1-Norm Semi-supervised Learning
This paper presents a novel L1-norm semisupervised learning algorithm for robust image analysis by giving new L1-norm formulation of Laplacian regularization which is the key step of graph-based semi-supervised learning. Since our L1-norm Laplacian regularization is defined directly over the eigenvectors of the normalized Laplacian matrix, we successfully formulate semi-supervised learning as a...
متن کاملArbitrary Norm Support Vector Machines
Support vector machines (SVM) are state-of-the-art classifiers. Typically L2-norm or L1-norm is adopted as a regularization term in SVMs, while other norm-based SVMs, for example, the L0-norm SVM or even the L(infinity)-norm SVM, are rarely seen in the literature. The major reason is that L0-norm describes a discontinuous and nonconvex term, leading to a combinatorially NP-hard optimization pro...
متن کاملAn Efficient Sparse Representation Algorithm for Direction-of-Arrival Estimation
This paper presents an efficient sparse representation approach to direction-of-arrival (DOA) estimation using uniform linear arrays. The proposed approach constructs the jointly sparse model in real domain by exploiting the properties of centro-Hermitian matrices. Subsequently, DOA estimation is realized via the sparse Bayesian learning (SBL) algorithm. Further, the pruning threshold of SBL is...
متن کاملBlind channel identification for speech dereverberation using l1-norm sparse learning
Speech dereverberation remains an open problem after more than three decades of research. The most challenging step in speech dereverberation is blind channel identification (BCI). Although many BCI approaches have been developed, their performance is still far from satisfactory for practical applications. The main difficulty in BCI lies in finding an appropriate acoustic model, which not only ...
متن کامل